What HMMs Can Do

نویسنده

  • Jeff A. Bilmes
چکیده

Since their inception almost fifty years ago, hidden Markov models (HMMs) have have become the predominant methodology for automatic speech recognition (ASR) systems — today, most state-of-the-art speech systems are HMM-based. There have been a number of ways to explain HMMs and to list their capabilities, each of these ways having both advantages and disadvantages. In an effort to better understand what HMMs can do, this tutorial article analyzes HMMs by exploring a definition of HMMs in terms of random variables and conditional independence assumptions. We prefer this definition as it allows us to reason more throughly about the capabilities of HMMs. In particular, it is possible to deduce that there are, in theory at least, no limitations to the class of probability distributions representable by HMMs. This paper concludes that, in search of a model to supersede the HMM (say for ASR), rather than trying to correct for HMM limitations in the general case, new models should be found based on their potential for better parsimony, computational requirements, and noise insensitivity. key words: Automatic Speech Recognition, Hidden Markov Models, HMMs, time-series processes, hand-writing recognition, graphical models, dynamic Bayesian networks, dynamic graphical models, stochastic processes, time-series densities, bio-informatics

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

What Hmms Can’t Do

Hidden Markov models (HMMs) are the predominant methodology for automatic speech recognition (ASR) systems. Ever since their inception, it has been said that HMMs are an inadequate statistical model for such purposes. Results over the years have shown, however, that HMM-based ASR performance continually improves given enough training data and engineering effort. In this paper, we argue that the...

متن کامل

Reviving discrete HMMs: the myth about the superiority of continuous HMMs

Despite what is generally believed, we have recently shown that discrete-distribution HMMs can outperform continuousdensity HMMs at significantly faster decoding speeds. Recognition performance and decoding speed of the discrete HMMs are improved by using product-code Vector Quantization (VQ) and mixtures of discrete distributions. In this paper, we present efficient training and decoding algor...

متن کامل

On the equivalence of Gaussian and log-linear HMMs

The acoustic models of conventional state-of-the-art speech recognition systems use generative Gaussian HMMs. In the past few years, discriminative models like for example Conditional Random Fields (CRFs) have been proposed to refine the acoustic models. CRFs directly model the class posteriors, the quantities of interest in recognition. CRFs are undirected models, and CRFs do not assume local ...

متن کامل

A Tutorial on Hidden Markov Models

In this tutorial we present an overview of i what are HMMs ii what are the di erent problems associated with HMMs iii the Viterbi algorithm for determining the optimal state sequence iv algorithms associated with training HMMs and v distance between HMMs

متن کامل

HMMs and Vision: Representing Structure and Sequences for Active Vision Using Hidden Markov Models

Hidden Markov models (HMMs) have rarely been used in computer vision, although they are a key part of modern speech recognition systems. We show how HMMs can be used in a variety of vision related tasks. In all cases, the HMM learns structured knowledge about the world from a set of observation sequences. The HMM can also be used to generate behavior. We present three examples. • An aspect grap...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • IEICE Transactions

دوره 89-D  شماره 

صفحات  -

تاریخ انتشار 2006